An evaluation of course evaluations
نویسندگان
چکیده
Student ratings of teaching have been used, studied, and debated for almost a century. This article examines student ratings of teaching from a statistical perspective. The common practice of relying on averages of student teaching evaluation scores as the primary measure of teaching effectiveness for promotion and tenure decisions should be abandoned for substantive and statistical reasons: There is strong evidence that student responses to questions of “effectiveness” do not measure teaching effectiveness. Response rates and response variability matter. And comparing averages of categorical responses, even if the categories are represented by numbers, makes little sense. Student ratings of teaching are valuable when they ask the right questions, report response rates and score distributions, and are balanced by a variety of other sources and methods to evaluate teaching. Since 1975, course evaluations at University of California, Berkeley, have asked: Considering both the limitations and possibilities of the subject matter and course, how would you rate the overall teaching effectiveness of this instructor? 1 (not at all effective), 2, 3, 4 (moderately effective), 5, 6, 7 (extremely effective) Among faculty, student evaluations of teaching (SET) are a source of pride and satisfaction—and frustration and anxiety. High-stakes decisions including tenure and promotions rely on SET. Yet it is widely believed that they are primarily a popularity contest, that it is easy to “game” ratings, that good teachers get bad ratings and vice versa, and that rating anxiety stifles pedagogical innovation and encourages faculty to water down course content. What is the truth? We review statistical issues in analyzing and comparing SET scores, problems defining and measuring teaching effectiveness, and pernicious distortions that result from using SET scores as a proxy for teaching quality and effectiveness. We argue here—and the literature shows—that students are in a good position to evaluate some aspects of teaching, but SET are at best tenuously connected to teaching effectiveness (defining and measuring teaching effectiveness are knotty problems in themselves; we discuss this below). Other ways of evaluating teaching can be combined with student comments to produce a more reliable and meaningful composite. We make recommendations regarding the use of SET and discuss new policies implemented at University of California, Berkeley, in 2013.
منابع مشابه
A Report on IAEA/RCA C-7RAS 004-061/6 Training Course in Chiba, Japan in 2014
The C7-RAS 6/061-004 training course by the International Atomic Energy Agency/Regional Cooperative Agreement (IAEA/RCA) was held in Chiba in 2014. The syllabus, pre- and post-course evaluations, and survey questionnaire results were assembled in this course. The post-course evaluation, including 32 questions similar to the pre-course evaluation, was performed right after the end of the final...
متن کاملSmartphone-based evaluations of clinical placements—a useful complement to web-based evaluation tools
PURPOSE Web-based questionnaires are currently the standard method for course evaluations. The high rate of smartphone adoption in Sweden makes possible a range of new uses, including course evaluation. This study examines the potential advantages and disadvantages of using a smartphone app as a complement to web-based course evaluationsystems. METHODS An iPhone app for course evaluations was...
متن کاملFudging the Numbers: Distributing Chocolate Influences Student Evaluations of an Undergraduate Course
Student evaluations provide important information about teaching effectiveness. Research has shown that student evaluations can be mediated by unintended aspects of a course. In this study, we examined whether an event unrelated to a course would increase student evaluations. Six discussion sections completed course evaluations administered by an independent experimenter. The experimenter offer...
متن کاملImpact of medical students' descriptive evaluations on long-term course development
BACKGROUND In medical education, feedback from students is helpful in course evaluation. However, the impact of medical students' feedback on long-term course development is seldom reported. In this project we studied the correspondence between medical students' descriptive evaluations and key features of course development over five years. METHODS Qualitative content analysis was used. The c...
متن کاملEvaluation of an evaluation
Introduction. Evaluation is a systematic way to improve and make more effective actions that involves procedures which are useful, feasible, ethical, and accurate. Common questions in all evaluations are: do all part of program do well and effective? What is the good functioning? Why the program or its parts do not work well? What are the effects and consequences of the program? Is this progra...
متن کاملCooperative learning and peer evaluation: The effect of free riders on team performance and the relationship between course performance and peer evaluation
Cooperative learning has gained popularity in higher educational settings. However, assigning grades equitably to all team members in a way that rewards them for their contributions remains challenging. In this paper, we ask whether having free riders on a team lowers the quality of submitted work, and whether students’ course performance correlates with peer evaluations. In an introductory soc...
متن کامل